Bayesian Optimization Using Sequential Monte Carlo

نویسندگان

  • Romain Benassi
  • Julien Bect
  • Emmanuel Vázquez
چکیده

We consider the problem of optimizing a real-valued continuous function f using a Bayesian approach, where the evaluations of f are chosen sequentially by combining prior information about f , which is described by a random process model, and past evaluation results. The main difficulty with this approach is to be able to compute the posterior distributions of quantities of interest which are used to choose evaluation points. In this article, we decide to use a Sequential Monte Carlo (SMC) approach. 1 Overview of the contribution proposed We consider the problem of finding the global maxima of a function f : X → R, where X ⊂ R is assumed bounded, using the expected improvement (EI) criterion [1, 3]. Many examples in the literature show that the EI algorithm is particularly interesting for dealing with the optimization of functions which are expensive to evaluate, as is often the case in design and analysis of computer experiments [2]. However, going from the general framework expressed in [1] to an actual computer implementation is a difficult issue. The main idea of an EI-based algorithm is a Bayesian one: f is viewed as a sample path of a random process ξ defined on R. For the sake of tractability, it is generally assumed that ξ has a Gaussian process distribution conditionally to a parameter θ ∈ Θ ⊆ R, which tunes the mean and covariance functions of the process. Then, given a prior distribution π0 on θ and some initial evaluation results ξ(X1), . . . , ξ(Xn0) at X1, . . . , Xn0 , an (idealized) EI algorithm constructs a sequence of evaluations points Xn0+1, Xn0+2, . . . such that, for each n ≥ n0, Xn+1 = argmax x∈X ρ̄n := ∫ θ∈Θ ρn(x; θ)dπn(θ) , (1) where πn stands for the posterior distribution of θ, conditional on the σ-algebra Fn generated by X1, ξ(X1), . . . , Xn, ξ(Xn), and ρn(x; θ) := En,θ((ξ(Xn+1)−Mn)+ | Xn+1 = x) is the EI at x given θ, with Mn = ξ(X0) ∨ · · · ∨ ξ(Xn) and En,θ the conditional expectation given Fn and θ. In practice, the computation of ρn is easily carried out (see [3]) but the answers to the following two questions will probably have a direct impact on the performance and applicability of a particular implementation: a) How to deal with the integral in ρ̄n? b) How to deal with the maximization of ρ̄n at each step? We can safely say that most implementations—including the popular EGO algorithm [3]—deal with the first issue by using an empirical Bayes (or plug-in) approach, which consists in approximating πn by a Dirac mass at the maximum likelihood estimate of θ. A plug-in approach using maximum a posteriori estimation has been used in [6]; fully Bayesian methods are more difficult to implement (see [4] and references therein). Regarding the optimization of ρ̄n at each step, several strategies have been proposed (see, e.g., [3, 5, 7, 10]). This article addresses both questions simultaneously, using a sequential Monte Carlo (SMC) approach [8, 9] and taking particular care to control the numerical complexity of the algorithm. The main ideas are the following. First, as in [5], a weighted sample Tn = {(θn,i, wn,i) ∈ Θ × R, 1 ≤ i ≤ I} from πn is used to approximate ρ̄n; that is, ∑I i=1 wn,i ρn(x; θn,i) →I ρ̄n(x). Besides, at each step n, we attach to each θn,i a (small) population of candidate evaluation points {xn,i,j , 1 ≤ j ≤ J} which is expected to cover promising regions for that particular value of θ and such that maxi,j ρ̄n (xn,i,j) ≈ maxx ρ̄n(x). 2 Algorithm and results At each step n ≥ n0 of the algorithm, our objective is to construct a set of weighted particles Gn = { ( γn,i,j , w ′ n,i,j ) , γn,i,j = (θn,i, xn,i,j) ∈ Θ ×X, w ′ n,i,j ∈ R , 1 ≤ i ≤ I, 1 ≤ j ≤ J }

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Sequential Monte Carlo Samplers

In this paper, we propose a methodology to sample sequentially from a sequence of probability distributions known up to a normalizing constant and defined on a common space. These probability distributions are approximated by a cloud of weighted random samples which are propagated over time using Sequential Monte Carlo methods. This methodology allows us to derive simple algorithms to make para...

متن کامل

Sequential Bayesian Optimisation for Spatial-Temporal Monitoring

Determine a non-myopic solution to the sequential decision making problem of monitoring and optimising a space and time dependent function using a moving sensor. Contributions: Sequential Bayesian Optimisation (SBO) Formulate SBO as a Partially Observed Markov Decision Process (POMDP). Find non-mypic solution for the POMDP analog of SBO using MonteCarlo Tree Search (MCTS) and Upper Confidence B...

متن کامل

Sequential Monte Carlo samplers for Bayesian DSGE models

Bayesian estimation of DSGE models typically uses Markov chain Monte Carlo as importance sampling (IS) algorithms have a difficult time in high-dimensional spaces. I develop improved IS algorithms for DSGE models using recent advances in Monte Carlo methods known as sequential Monte Carlo samplers. Sequential Monte Carlo samplers are a generalization of particle filtering designed for full simu...

متن کامل

Generalised linear mixed model analysis via sequential Monte Carlo sampling

We present a sequential Monte Carlo sampler algorithm for the Bayesian analysis of generalised linear mixed models (GLMMs). These models support a variety of interesting regression-type analyses, but performing inference is often extremely difficult, even when using the Bayesian approach combined with Markov chain Monte Carlo (MCMC). The Sequential Monte Carlo sampler (SMC) is a new and general...

متن کامل

A sequential Monte Carlo approach to Thompson sampling for Bayesian optimization

This paper first presents a novel algorithm approximating the distribution of the maximum (both its position and its value) of a Gaussian process. This algorithm uses particles in a similar way as Sequential Monte Carlo samplers. It is subsequently applied to the problem of Gaussian Process Optimization (GPO). The resulting GPO algorithm does not use an acquisition function, which makes it diff...

متن کامل

Adaptive Path Design of a Moving Radar

We consider the problem of designing the trajectory of a radar mounted on a moving platform. We develop an adaptive algorithm that, at each time step, optimally selects the radar path in response to the estimated and predicted target parameters to improve the tracking accuracy. We derive our approach under a framework of sequential Bayesian filtering. We apply a sequential Monte Carlo method (p...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2012